Euclidean Embeddings that Preserve Volumes

نویسنده

  • Anastasios Zouzias
چکیده

Let P be a set of n points in Euclidean space and let 0 < ε < 1. A well-known result of Johnson and Lindenstrauss states that there is a projection of P onto a subspace of dimension O(ε−2 logn) such that distances change by a factor of 1 + ε at most. We consider an extension of this result. Our goal is to find an analogous dimension reduction where not only pairs, but all subsets of at most k points maintain their volume approximately. More precisely, we require that sets of size s ≤ k preserve their volumes within a factor of (1 + ε)s−1. We show that this can be achieved using O(max{k/ε,ε−2 logn}) dimensions. This in particular means that for k = O(logn/ε) we require no more dimensions (asymptotically) than the special case k = 2, handled by Johnson and Lindenstrauss. Our work improves on a result of Magen that required as many as O(kε−2 logn) dimensions and is tight up to a factor of O(1/ε). Another outcome of our work is an alternative and simplified proof of the result of Magen showing that all distances between points and affine subspaces spanned by a small number of points are approximately preserved when projecting onto O(kε−2 logn) dimensions. Instead of fixing the distortion and finding the required dimension, we also consider

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Hyperbolic Entailment Cones for Learning Hierarchical Embeddings

Learning graph representations via lowdimensional embeddings that preserve relevant network properties is an important class of problems in machine learning. We here present a novel method to embed directed acyclic graphs. Following prior work, we first advocate for using hyperbolic spaces which provably model tree-like structures better than Euclidean geometry. Second, we view hierarchical rel...

متن کامل

Volume in General Metric Spaces

A central question in the geometry of finite metric spaces is how well can an arbitrary metric space be “faithfully preserved” by a mapping into Euclidean space. In this paper we present an algorithmic embedding which obtains a new strong measure of faithful preservation: not only does it (approximately) preserve distances between pairs of points, but also the volume of any set of k points. Suc...

متن کامل

Word Re-Embedding via Manifold Dimensionality Retention

Word embeddings seek to recover a Euclidean metric space by mapping words into vectors, starting from words cooccurrences in a corpus. Word embeddings may underestimate the similarity between nearby words, and overestimate it between distant words in the Euclidean metric space. In this paper, we re-embed pre-trained word embeddings with a stage of manifold learning which retains dimensionality....

متن کامل

Hybed: Hyperbolic Neural Graph Embedding

Neural embeddings have been used with great success in Natural Language Processing (NLP). They provide compact representations that encapsulate word similarity and attain state-of-the-art performance in a range of linguistic tasks. The success of neural embeddings has prompted significant amounts of research into applications in domains other than language. One such domain is graph-structured d...

متن کامل

A note on random projections for preserving paths on a manifold

Random projections are typically used to study low distortion linear embeddings that approximately preserve Euclidean distances between pairs of points in a set S ⊂ R. Of particular interest is when the set S is a low-dimensional submanifold of R. Recent results by Baraniuk and Wakin [2007] and Clarkson [2007] shed light on how to pick the projection dimension to achieve low distortion of Eucli...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009